Hello everyone, I am Stefan, starting today to bring you a detailed Hadoop learning tutorial, you can follow my tutorial step by step into the development of cloud computing, OK, nonsense, we started the first: Hadoop environment.
The beginning of everything is difficult, this is not a blow. Many people in the initial environment to build up the problem, and everyone's platform and there are differences, it
Hadoop Modes
Pre-install Setup
Creating a user
SSH Setup
Installing Java
Install Hadoop
Install in Standalone Mode
Lets do a test
Install in Pseudo distributed Mode
Hadoop
-connector-java-5.0.8/mysql-connector-java-5.0.8-bin.jar./libTo start hive:$ cd/home/zxm/hadoop/hive-0.8.1;./bin/hiveTest:$./hiveWARNING:org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please use org.apache.hadoop.log.metrics.EventCounter the log4j.properties files.Logging initialized using configuration in jar:file:/home/zxm/hadoop/hive-0.8.1/lib/hive-common-0.8.1.jar!/ Hive-log4j.propertiesHive
-1.2.1export PATH=$PATH:$HADOOP_HOME/binexport HADOOP_HOME_WARN_SUPPRESS=13) Make the configuration file effective[[emailprotected] ~]$ source /etc/profilefor more details, please read on to the next page. Highlights : http://www.linuxidc.com/Linux/2015-03/114669p2.htm--------------------------------------Split Line--------------------------------------Ubuntu14.04 Hadoop2.4.1 stand-alone/pseudo-distributed installation configuration tutorial http://www.linuxidc.com/Linux/2015-02/113487.htmCentOS
I. Create Hadoop user portfolio under Ubuntu Hadoop user1. Create a Hadoop user group addgroup HADOOP2, create a Hadoop user adduser-ingroup Hadoop hadoop3, Add permissions NBSP;VIM/ETC/SUDOERS4 to Hadoop users, switch to
original path to the target path Hadoop fs-cat/user/hadoop/a.txt View the contents of the A.txt file Hadoop fs-rm/user/hadoop/a.txt Delete US The A.txt file below the Hadoop folder under the ER folderHadoop fs-rm-r/user/hadoop/a.
completes the modification of the Hadoop-eclipse-plugin-0.20.203.0.jar.
Finally, copy the Hadoop-eclipse-plugin-0.20.203.0.jar to the plugins directory of Eclipse:
$ CD ~/hadoop-0.20.203.0/lib
$ sudo cp hadoop-eclipse-plugin-0.20.203.0.jar/usr/eclipse/plugins/
5. Configure the plug-in in Eclipse.
First, open Eclipse
I recently tried to build the environment for Hadoop, but I really don't know how to build it. The next hop was a step-by-step error. Answers from many people on the Internet are also common pitfalls (for example, the most typical is the case sensitivity of commands, for example, hadoop commands are in lower case, and many people write Hadoop, so when you encount
I. Create Hadoop user portfolio under Ubuntu Hadoop user1. Create a Hadoop user group addgroup HADOOP2, create a Hadoop user adduser-ingroup Hadoop hadoop3, Add permissions for Hadoop users vim/etc/sudoers 4, switch to
"1.7.0_79"Java (TM) SE Runtime Environment (build 1.7.0_79-b15)Java HotSpot (TM) Client VM (build 24.79-b02, Mixed mode)Indicates that the JDK environment variable is configured successfullyThird, install Hadoop3.1 Download Hadoop, choose Stable version, in fact stable version is 1.2.1, download the site as follows:Http://mirror.esocc.com/apache/hadoop/common/hadoop
comment #) Note: Some blogs write that you need to comment out the next line
export hadoop_opts= "-djava.security.krb5.realm=ox. ac.uk-djava.security.krb5.kdc=kdc0.ox.ac.uk:kdc1.ox.ac.uk "(remove comments) I didn't find this one, so I didn't have this one.
2. Configuration core-site.xml--Specifies the hostname and port of the Namenode
4. Configuration mapred-site.xml--Specifies the hostname and port of the Jobtracker
5.SSH configuration turn on sharing in
Setup function source code: (Excerpt from "Hadoop Combat")*called once at the start of the task.protected void Setup (context context) throws ioexception,interruptedexception{}As you can tell from the comments, the setup function is called when the task starts.Jobs in MapReduce are organized into Maptask and Reducetask
The main process for installing and setting up Hadoop under Ubuntu.1. Create a Hadoop userCreate a user named Hadoop and create the user's home directory under home without detailed description.2. Installing the Java EnvironmentDownload the jdk:jdk-8u111-linux-x64.tar.gz under Linux environment.Create a Java folder under USR, copy the jdk-8u111-linux-x64.tar.gz t
file./hdfs/data--Storing data./hdfs/tmp--Storing temporary files 2.6 Modifying an XML configuration file The XML file that needs to be modified under hadoop2.9.1/etc/hadoop/There are 5 main files to modify:hadoop-env.shCore-site.xmlHdfs-site.xmlMapred-site.xmlYarn-site.xmlSlaves 2.6.1, vim hadoop-env.sh, fill in the Java installation path 2.6.2, vim core-site.xml,configuration tag insert t
Apache Hadoop2.2.0, as the next-generation hadoop version, breaks through the limit of up to 4000 machines in the original hadoop1.x cluster, and effectively solves the frequently encountered OOM (memory overflow) problem, its innovative computing framework, YARN, is called the hadoop operating system. It is not only compatible with the original mapreduce computing model, but also supports other parallel co
The role of Setup run cleanup context in Hadoop execution1. Introduction1) Setup (), this method is only executed once by the MapReduce framework and performs a centralized initialization of the relevant variables or resources before performing the map task. If the resource initialization work is placed in the method map (), causing the mapper task to parse each
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.